15 research outputs found

    Prediction and control of nonlinear dynamical systems using machine learning

    Get PDF
    Künstliche Intelligenz und Machine Learning erfreuen sich in Folge der rapide gestiegenen Rechenleistung immer größerer Popularität. Sei es autonomes Fahren, Gesichtserkennung, bildgebende Diagnostik in der Medizin oder Robotik – die Anwendungsvielfalt scheint keine Grenzen zu kennen. Um jedoch systematischen Bias und irreführende Ergebnisse zu vermeiden, ist ein tiefes Verständnis der Methoden und ihrer Sensitivitäten von Nöten. Anhand der Vorhersage chaotischer Systeme mit Reservoir Computing – einem künstlichen rekurrenten neuronalem Netzwerk – wird im Rahmen dieser Dissertation beleuchtet, wie sich verschiedene Eigenschaften des Netzwerks auf die Vorhersagekraft und Robustheit auswirken. Es wird gezeigt, wie sich die Variabilität der Vorhersagen – sowohl was die exakte zukünftige Trajektorie betrifft als auch das statistische Langzeitverhalten (das "Klima") des Systems – mit geeigneter Parameterwahl signifikant reduzieren lässt. Die Nichtlinearität der Aktivierungsfunktion spielt hierbei eine besondere Rolle, weshalb ein Skalierungsparameter eingeführt wird, um diese zu kontrollieren. Des Weiteren werden differenzielle Eigenschaften des Netzwerkes untersucht und gezeigt, wie ein kontrolliertes Entfernen der "richtigen" Knoten im Netzwerk zu besseren Vorhersagen führt und die Größe des Netzwerkes stark reduziert werden kann bei gleichzeitig nur moderater Verschlechterung der Ergebnisse. Dies ist für Hardware Realisierungen von Reservoir Computing wie zum Beispiel Neuromorphic Computing relevant, wo möglichst kleine Netzwerke von Vorteil sind. Zusätzlich werden unterschiedliche Netzwerktopologien wie Small World Netzwerke und skalenfreie Netzwerke beleuchtet. Mit den daraus gewonnenen Erkenntnissen für bessere Vorhersagen von nichtlinearen dynamischen Systemen wird eine neue Kontrollmethode entworfen, die es ermöglicht, dynamische Systeme flexibel in verschiedenste Zielzustände zu lenken. Hierfür wird – anders als bei vielen bisherigen Ansätzen – keine Kenntnis der zugrundeliegenden Gleichungen des Systems erfordert. Ebenso wird nur eine begrenzte Datenmenge verlangt, um Reservoir Computing hinreichend zu trainieren. Zudem ist es nicht nur möglich, chaotisches Verhalten in einen periodischen Zustand zu zwingen, sondern auch eine Kontrolle auf komplexere Zielzustände wie intermittentes Verhalten oder ein spezifischer anderer chaotischer Zustand. Dies ermöglicht eine Vielzahl neuer potenzieller realer Anwendungen, von personalisierten Herzschrittmachern bis hin zu Kontrollvorrichtungen für Raketentriebwerke zur Unterbindung von kritischen Verbrennungsinstabilitäten. Als Schritt zur Weiterentwicklung von Reservoir Computing zu einem verbesserten hybriden System, das nicht nur rein datenbasiert arbeitet, sondern auch physikalische Zusammenhänge berücksichtigt, wird ein Ansatz vorgestellt, um lineare und nichtlinearen Kausalitätsstrukturen zu separieren. Dies kann verwendet werden, um Systemgleichungen oder Restriktionen für ein hybrides System zur Vorhersage oder Kontrolle abzuleiten.Artificial intelligence and machine learning are becoming increasingly popular as a result of the rapid increase in computing power. Be it autonomous driving, facial recognition, medical imaging diagnostics or robotics – the variety of applications seems to have no limits. However, to avoid systematic bias and misleading results, a deep understanding of the methods and their sensitivities is needed. Based on the prediction of chaotic systems with reservoir computing – an artificial recurrent neural network – this dissertation sheds light on how different properties of the network affect the predictive power and robustness. It is shown how the variability of the predictions – both in terms of the exact short-term predictions and the long-term statistical behaviour (the "climate") of the system – can be significantly reduced with appropriate parameter choices. The nonlinearity of the activation function plays a special role here, thus a scaling parameter is introduced to control it. Furthermore, differential properties of the network are investigated and it is shown how a controlled removal of the right nodes in the network leads to better predictions, whereas the size of the network can be greatly reduced while only moderately degrading the results. This is relevant for hardware realizations of reservoir computing such as neuromorphic computing, where networks that are as small as possible are advantageous. Additionally, different network topologies such as small world networks and scale-free networks are investigated. With the insights gained for better predictions of nonlinear dynamical systems, a new control method is designed that allows dynamical systems to be flexibly forced into a wide variety of dynamical target states. For this – unlike many previous approaches – no knowledge of the underlying equations of the system is required. Further, only a limited amount of data is needed to sufficiently train reservoir computing. Moreover, it is possible not only to force chaotic behavior to a periodic state, but also to control for more complex target states such as intermittent behavior or a specific different chaotic state. This enables a variety of new potential real-world applications, from personalized cardiac pacemakers to control devices for rocket engines to suppress critical combustion instabilities. As a step toward advancing reservoir computing to an improved hybrid system that is not only purely data-based but also takes into account physical relationships, an approach is presented to separate linear and nonlinear causality structures. This can be used to derive system equations or constraints for a hybrid prediction or control system

    Linear and nonlinear market correlations: characterizing financial crises and portfolio optimization

    Get PDF
    Pearson correlation and mutual information based complex networks of the day-to-day returns of US S&P500 stocks between 1985 and 2015 have been constructed in order to investigate the mutual dependencies of the stocks and their nature. We show that both networks detect qualitative differences especially during (recent) turbulent market periods thus indicating strongly fluctuating interconnections between the stocks of different companies in changing economic environments. A measure for the strength of nonlinear dependencies is derived using surrogate data and leads to interesting observations during periods of financial market crises. In contrast to the expectation that dependencies reduce mainly to linear correlations during crises we show that (at least in the 2008 crisis) nonlinear effects are significantly increasing. It turns out that the concept of centrality within a network could potentially be used as some kind of an early warning indicator for abnormal market behavior as we demonstrate with the example of the 2008 subprime mortgage crisis. Finally, we apply a Markowitz mean variance portfolio optimization and integrate the measure of nonlinear dependencies to scale the investment exposure. This leads to significant outperformance as compared to a fully invested portfolio.Comment: 12 pages, 11 figures, Phys. Rev. E, accepte

    Controlling dynamical systems to complex target states using machine learning: next-generation vs. classical reservoir computing

    Full text link
    Controlling nonlinear dynamical systems using machine learning allows to not only drive systems into simple behavior like periodicity but also to more complex arbitrary dynamics. For this, it is crucial that a machine learning system can be trained to reproduce the target dynamics sufficiently well. On the example of forcing a chaotic parametrization of the Lorenz system into intermittent dynamics, we show first that classical reservoir computing excels at this task. In a next step, we compare those results based on different amounts of training data to an alternative setup, where next-generation reservoir computing is used instead. It turns out that while delivering comparable performance for usual amounts of training data, next-generation RC significantly outperforms in situations where only very limited data is available. This opens even further practical control applications in real world problems where data is restricted.Comment: IJCNN 202

    Controlling nonlinear dynamical systems into arbitrary states using machine learning

    Get PDF
    Controlling nonlinear dynamical systems is a central task in many different areas of science and engineering. Chaotic systems can be stabilized (or chaotified) with small perturbations, yet existing approaches either require knowledge about the underlying system equations or large data sets as they rely on phase space methods. In this work we propose a novel and fully data driven scheme relying on machine learning (ML), which generalizes control techniques of chaotic systems without requiring a mathematical model for its dynamics. Exploiting recently developed ML-based prediction capabilities, we demonstrate that nonlinear systems can be forced to stay in arbitrary dynamical target states coming from any initial state. We outline and validate our approach using the examples of the Lorenz and the Rössler system and show how these systems can very accurately be brought not only to periodic, but even to intermittent and different chaotic behavior. Having this highly flexible control scheme with little demands on the amount of required data on hand, we briefly discuss possible applications ranging from engineering to medicine

    Reducing network size and improving prediction stability of reservoir computing

    Get PDF
    Reservoir computing is a very promising approach for the prediction of complex nonlinear dynamical systems. Besides capturing the exact short-term trajectories of nonlinear systems, it has also proved to reproduce its characteristic long-term properties very accurately. However, predictions do not always work equivalently well. It has been shown that both short- and long-term predictions vary significantly among different random realizations of the reservoir. In order to gain an understanding on when reservoir computing works best, we investigate some differential properties of the respective realization of the reservoir in a systematic way. We find that removing nodes that correspond to the largest weights in the output regression matrix reduces outliers and improves overall prediction quality. Moreover, this allows to effectively reduce the network size and, therefore, increase computational efficiency. In addition, we use a nonlinear scaling factor in the hyperbolic tangent of the activation function. This adjusts the response of the activation function to the range of values of the input variables of the nodes. As a consequence, this reduces the number of outliers significantly and increases both the short- and long-term prediction quality for the nonlinear systems investigated in this study. Our results demonstrate that a large optimization potential lies in the systematical refinement of the differential reservoir properties for a given dataset.Comment: 11 pages, 8 figures, published in Chao

    Evaluation of the South Asian Monsoon in the ECHAM/MESSy

    No full text
    The South Asian Monsoon has been evaluated in the ECHAM/MESSy Atmo-spheric Chemistry (EMAC) model in comparison to observations and models par-ticipating in the 5th Phase of the Coupled Model Intercomparison Project (CMIP5).Existing diagnostics of the Earth System Model Evaluation Tool (ESMValTool)have been applied to historical CMIP5 simulations and to an EMAC timeslice ex-periment representing the year 2000 to investigate precipitation in the South AsiaMonsoon period. The EMAC simulation generally overestimates precipitation inSouth Asia during the local summer and winter seasons compared to observations,and also the global precipitation intensity, which is calculated as summer minuswinter difference. This bias in precipitation intensity results in an overestimationof the global monsoon domains, which are areas that exceed a precipitation inten-sity of 2.5 mm per day. Compared to the CMIP5 historical simulations, not only theCMIP5 multi-model mean but also each individual model is more skillful in termsof pattern correlation index. Possible reasons for biases in the EMAC simulationcould for example arise from biases in circulation or the representation of clouds.The simulation of the 850 hPa low-level wind was better represented than the pre-cipitation intensity, although some biases compared to meteorological reanalysisexist. Diagnostics have been newly developed and implemented into the ESMVal-Tool to examine the representation of Cloud Radiative Effects (CRE) as externalsolar radiation is the main driver for monsoon. Deviations from observations arefound in particular for the simulated shortwave radiation, which could be relatedto problems in the representation of low-level clouds. The biases found in theparticular EMAC simulation evaluated here could partly be due to the boundaryconditions used, in particular the prescribed sea surface temperatures (SST) thatare taken from a CMIP5 model simulation with the CMCC model. In addition, acoupled ocean could possibly partially compensate for the biases in precipitation.Additional work is required to investigate this further. For example, an EMACsimulation with observed SSTs or a coupled ocean could be evaluated to betterunderstand the reasons for the biases found her

    Good and bad predictions: Assessing and improving the replication of chaotic attractors by means of reservoir computing

    Get PDF
    The prediction of complex nonlinear dynamical systems with the help of machine learning techniques has become increasingly popular. In particular, reservoir computing turned out to be a very promising approach especially for the reproduction of the long-term properties of a nonlinear system. Yet, a thorough statistical analysis of the forecast results is missing. Using the Lorenz and Rössler system, we statistically analyze the quality of prediction for different parametrizations - both the exact short-term prediction as well as the reproduction of the longterm properties (the climate) of the system as estimated by the correlation dimension and largest Lyapunov exponent. We find that both short- and long-term predictions vary significantly among the realizations. Thus, special care must be taken in selecting the good predictions as realizations, which deliver better short-term prediction also tend to better resemble the long-term climate of the system. Instead of only using purely random Erdös-Renyi networks, we also investigate the benefit of alternative network topologies such as small world or scalefree networks and show which effect they have on the prediction quality. Our results suggest that the overall performance with respect to the reproduction of the climate of both the Lorenz and Rössler system is worst for scale-free networks. For the Lorenz system, there seems to be a slight benefit of using small world networks, while for the Rössler system, small world and Erdös-Renyi networks performed equivalently well. In general, the observation is that reservoir computing works for all network topologies investigated here

    Complex network-based analysis of nonlinear dependencies in multidimensional financial time series

    Get PDF
    Cross-correlation and mutual information based complex networks of the day-to-day returns of US S&P500 stocks between 1985 and 2015 have been constructed in order to investigate the mutual dependencies of the stocks and their nature. We show that both networks detect qualitative differences especially during (recent) turbulent market periods thus indicating strongly fluctuating interconnections between the stocks of different companies in changing economic environments. A measure for the strength of nonlinear dependencies has been derived using surrogate data and led to interesting observations during periods of financial market crisis. In contrast to the prevailing view that dependencies reduce mainly to linear correlations during crisis it turned out that (at least in the crisis after 2008) nonlinear effects are significantly increasing. Finally, we apply a Markowitz mean variance portfolio optimization and integrate the measure of nonlinear dependencies to scale the investment exposure. This leads to significant outperformance as compared to a fully invested portfolio

    Controlling dynamical systems to complex target states using machine learning: next-generation vs. classical reservoir computing

    No full text
    Controlling nonlinear dynamical systems using machine learning allows to not only drive systems into simple behavior like periodicity but also to more complex arbitrary dynamics. For this, it is crucial that a machine learning system can be trained to reproduce the target dynamics sufficiently well. On the example of forcing a chaotic parametrization of the Lorenz system into intermittent dynamics, we show first that classical reservoir computing excels at this task. In a next step, we compare those results based on different amounts of training data to an alternative setup, where next-generation reservoir computing is used instead. It turns out that while delivering comparable performance for usual amounts of training data, next-generation RC significantly outperforms in situations where only very limited data is available. This opens even further practical control applications in real world problems where data is restricted

    Nonlinear correlations in financial markets: A complex network based analysis

    No full text
    It is common practice in finance to quantify correlations among financial time series in terms of their linear Pearson correlation coefficient. Knowing that financial time series show intermittent behavior being reminiscent of turbulence and leading to the well-known fat tails in the probability distribution as well as nonlinearities that also significantly show up as deviations from randomness in the distribution of Fourier phases, it is justified to assume that also nonlinear correlations among financial time series may be present. Therefore, Pearson correlation and mutual information based complex networks of the day-to-day returns of US S&P500 stocks between 1985 and 2015 have been constructed in order to investigate the mutual dependencies of the stocks and their nature. By deriving a measure for the strength of nonlinear correlations using surrogate data we show that a significant amount of information is lost when only relying on linear correlations measures.Our studies revealed that in contrast to the expectation that dependencies reduce mainly to linear correlations during crises, at least in the 2008 crisis nonlinear effects are significantly increasing. More specifically, we find that there are different types of financial crises in terms of nonlinear effects. Our results indicate that during the 2008 crisis nonlinear effects were significantly stronger than in preceding crises like the bursting of the Dot-com bubble. Furthermore, we found that major political events seem to have no significant mid- and long-term impact on market correlation structure in contrast to financial crises. In addition, it turns out that the concept of centrality within a network could potentially be used as some kind of an early warning indicator for abnormal market behavior as we also demonstrate with the example of the 2008 subprime mortgage crisis. Finally we developed a practical application in the field of portfolio optimization. We showed that scaling the investment exposure based on the strength of nonlinear correlations leads to an outperformance compared to a fully invested portfolio with static exposure. Furthermore, we obtained first interesting results on the relationship of linear and nonlinear correlations with causality measure
    corecore